AIbase
Home
AI Tools
AI Models
MCP
AI NEWS
EN
Model Selection
Tags
AWQ quantization

# AWQ quantization

Deepseek R1 0528 AWQ
MIT
AWQ-quantized version of DeepSeek R1 0528, supports full-context-length operation on 8x80GB GPUs using vLLM.
Large Language Model Transformers Supports Multiple Languages
D
cognitivecomputations
145
7
Llama 3.3 70B Instruct AWQ
The AWQ 4-bit quantization version of Llama-3.3-70B-Instruct, developed by Meta AI, supports multiple languages and is suitable for text generation tasks.
Large Language Model Transformers Supports Multiple Languages
L
kosbu
379.28k
6
Dolphin 2.9 Llama3 70b Awq
AWQ quantized version of Dolphin 2.9 Llama3 70B, suitable for vllm and other inference engines.
Large Language Model Transformers
D
julep-ai
19
5
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
English简体中文繁體中文にほんご
© 2025AIbase